Robust Independence Testing for Constraint-Based Learning of Causal Structure
نویسندگان
چکیده
This paper considers a method that combines ideas from Bayesian learning, Bayesian network inference, and classical hypothesis testing to produce a more reliable and robust test of independence for constraintbased (CB) learning of causal structure. Our method produces a smoothed contingency table NXY Z that can be used with any test of independence that relies on contingency table statistics. NXY Z can be calculated in the same asymptotic time and space required to calculate a standard contingency table, allows the specification of a prior distribution over parameters, and can be calculated when the database is incomplete. We provide theoretical justification for the procedure, and with synthetic data we demonstrate its benefits empirically over both a CB algorithm using the standard contingency table, and over a greedy Bayesian algorithm. We show that, even when used with noninformative priors, it results in better recovery of structural features and it produces networks with smaller KL-Divergence, especially as the number of nodes increases or the number of records decreases. Another benefit is the dramatic reduction in the probability that a CB algorithm will stall during the search, providing a remedy for an annoying problem plaguing CB learning when the database is small.
منابع مشابه
A Robust Independence Test for Constraint-Based Learning of Causal Structure
Constraint-based (CB) learning is a formalism for learning a causal network with a database D by performing a series of conditionalindependence tests Std-IT (X,Y | Z, D) to infer structural information. This paper considers a new test of independence Bayes-IT (X, Y | Z, D) that combines ideas from Bayesian learning, Bayesian network inference, and classical hypothesis testing to produce a more ...
متن کاملDiscovery of Causal Models that Contain Latent Variables Through Bayesian Scoring of Independence Constraints
Discovering causal structure from observational data in the presence of latent variables remains an active research area. Constraint-based causal discovery algorithms are relatively efficient at discovering such causal models from data using independence tests. Typically, however, they derive and output only one such model. In contrast, Bayesian methods can generate and probabilistically score ...
متن کاملThe Research-Engaged School: The Development and Test of a Causal Model through an Exploratory Mixed Methods Design
The Research-Engaged School: The Development and Test of a Causal Model through an Exploratory Mixed Methods Design Sh. HosseinPour, Ph.D.[1] H.R. Zeinabadi, Ph.D. [2] The present study was undertaken to design and test a research-engaged school model using mixed methods design. In the qualitative and quantitative parts of the study, phenomenological strategy and structural equation mod...
متن کاملJoint Probabilistic Inference of Causal Structure
Causal directed acyclic graphical models (DAGs) are powerful reasoning tools in the study and estimation of cause and effect in scientific and socio-behavioral phenomena. In many domains where the cause and effect structure is unknown, a key challenge in studying causality with DAGs is learning the structure of causal graphs directly from observational data. Traditional approaches to causal str...
متن کاملPermutation Testing Improves Bayesian Network Learning
We are taking a peek “under the hood” of constraint-based learning of graphical models such as Bayesian Networks. This mainstream approach to learning is founded on performing statistical tests of conditional independence. In all prior work however, the tests employed for categorical data are only asymptotically-correct, i.e., they converge to the exact p-value in the sample limit. In the prese...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2003